|
In statistics, the Bayesian information criterion (BIC) or Schwarz criterion (also SBC, SBIC) is a criterion for model selection among a finite set of models; the model with the lowest BIC is preferred. It is based, in part, on the likelihood function and it is closely related to the Akaike information criterion (AIC). When fitting models, it is possible to increase the likelihood by adding parameters, but doing so may result in overfitting. Both BIC and AIC resolve this problem by introducing a penalty term for the number of parameters in the model; the penalty term is larger in BIC than in AIC. The BIC was developed by Gideon E. Schwarz and published in a 1978 paper,〔.〕 where he gave a Bayesian argument for adopting it. == Definition == The BIC is formally defined as : where * = the observed data; * = the parameters of the model; * = the number of data points in , the number of observations, or equivalently, the sample size; * = the number of free parameters to be estimated. If the model under consideration is a linear regression, is the number of regressors, including the intercept; * = the maximized value of the likelihood function of the model , i.e. , where are the parameter values that maximize the likelihood function. The BIC is an asymptotic result derived under the assumptions that the data distribution is in the exponential family. That is, the integral of the likelihood function times the prior probability distribution over the parameters of the model for fixed observed data is approximated as : For large , this can be approximated by the formula given above. The BIC is used in model selection problems where adding a constant to the BIC does not change the result. 抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)』 ■ウィキペディアで「Bayesian information criterion」の詳細全文を読む スポンサード リンク
|